2026-04-10 18:55:16,120 [ 316201 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2026-04-10 18:55:16,120 [ 316201 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:79, check_args_and_update_paths) 2026-04-10 18:55:16,120 [ 316201 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:90, check_args_and_update_paths) 2026-04-10 18:55:16,120 [ 316201 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:92, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_403iw5 --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server]' 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user]' 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server]' 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user]' test_overcommit_tracker/test.py::test_user_overcommit -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [5 items] scheduling tests via LoadFileScheduling test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server] test_overcommit_tracker/test.py::test_user_overcommit [gw0] [ 20%] SKIPPED test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server] test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user] [gw0] [ 40%] SKIPPED test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user] test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server] [gw0] [ 60%] SKIPPED test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server] test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user] [gw0] [ 80%] SKIPPED test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user] [gw2] [100%] FAILED test_overcommit_tracker/test.py::test_user_overcommit =================================== FAILURES =================================== _____________________________ test_user_overcommit _____________________________ [gw2] linux -- Python 3.10.12 /usr/bin/python3 def test_user_overcommit(): node.query("CREATE USER IF NOT EXISTS A") node.query("GRANT ALL ON *.* TO A") responses_A = list() responses_B = list() for i in range(100): if i % 2 == 0: responses_A.append(node.get_query_request(USER_TEST_QUERY_A, user="A")) else: responses_B.append(node.get_query_request(USER_TEST_QUERY_B, user="A")) overcommited_killed = False for response in responses_A: _, err = response.get_answer_and_error() if "MEMORY_LIMIT_EXCEEDED" in err: overcommited_killed = True finished = False for response in responses_B: _, err = response.get_answer_and_error() if err == "": finished = True > assert finished, "all tasks are killed" E AssertionError: all tasks are killed E assert False test_overcommit_tracker/test.py:52: AssertionError ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2026-04-10 18:55:22.807000 [ 638 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:55:22.826000 [ 638 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:55:22.826000 [ 638 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2026-04-10 18:55:22.826000 [ 638 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2026-04-10 18:55:22.827000 [ 638 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2026-04-10 18:55:22.846000 [ 638 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2026-04-10 18:55:22.850000 [ 638 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2026-04-10 18:55:22.850000 [ 638 ] INFO : Running tests in /ClickHouse/tests/integration/test_overcommit_tracker/test.py (cluster.py:2748, start) 2026-04-10 18:55:22.851000 [ 638 ] DEBUG : Cluster start called. is_up=False (cluster.py:2755, start) 2026-04-10 18:55:22.869000 [ 638 ] DEBUG : Docker networks for project roottestovercommittracker-gw2 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:55:22.890000 [ 638 ] DEBUG : Docker containers for project roottestovercommittracker-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:55:22.912000 [ 638 ] DEBUG : Docker volumes for project roottestovercommittracker-gw2 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:55:22.912000 [ 638 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 18:55:22.934000 [ 638 ] DEBUG : Docker networks for project roottestovercommittracker-gw2 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:55:22.953000 [ 638 ] DEBUG : Docker containers for project roottestovercommittracker-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:55:22.972000 [ 638 ] DEBUG : Docker volumes for project roottestovercommittracker-gw2 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:55:22.972000 [ 638 ] DEBUG : Command:[docker container list --all --filter name='^/roottestovercommittracker-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 18:55:22.992000 [ 638 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 18:55:22.992000 [ 638 ] DEBUG : No running containers for project: roottestovercommittracker-gw2 (cluster.py:889, cleanup) 2026-04-10 18:55:22.992000 [ 638 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 18:55:23.010000 [ 638 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 18:55:23.011000 [ 638 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 18:55:23.038000 [ 638 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 18:55:23.038000 [ 638 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 18:55:23.039000 [ 638 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 18:55:23.039000 [ 638 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:55:23.056000 [ 638 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:55:23.057000 [ 638 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) 2026-04-10 18:55:23.057000 [ 638 ] DEBUG : Setup directory for instance: node (cluster.py:2768, start) 2026-04-10 18:55:23.057000 [ 638 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 18:55:23.057000 [ 638 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 18:55:23.058000 [ 638 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 18:55:23.058000 [ 638 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 18:55:23.059000 [ 638 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_overcommit_tracker/configs/config.d/config.xml'] to /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 18:55:23.059000 [ 638 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/database (cluster.py:4772, create_dir) 2026-04-10 18:55:23.059000 [ 638 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/logs (cluster.py:4783, create_dir) 2026-04-10 18:55:23.059000 [ 638 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2026-04-10 18:55:23.060000 [ 638 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'LSAN_OPTIONS': 'suppressions=/etc/clickhouse-server/lsan_suppressions.txt', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/.env (cluster.py:96, _create_env_file) 2026-04-10 18:55:23.060000 [ 638 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 18:55:23.060000 [ 638 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 18:55:23.060000 [ 638 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 18:55:23.060000 [ 638 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 18:55:23.069000 [ 638 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2026-04-10 18:55:23.070000 [ 638 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/.env --project-name roottestovercommittracker-gw2 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/docker-compose.yml pull] (cluster.py:121, run_and_check) 2026-04-10 18:55:33.491000 [ 638 ] DEBUG : Stderr: node Pulling (cluster.py:147, run_and_check) 2026-04-10 18:55:33.491000 [ 638 ] DEBUG : Stderr: node Pulled (cluster.py:147, run_and_check) 2026-04-10 18:55:33.491000 [ 638 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/.env --project-name roottestovercommittracker-gw2 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/docker-compose.yml up -d --no-recreate') (cluster.py:3149, start) 2026-04-10 18:55:33.491000 [ 638 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/.env --project-name roottestovercommittracker-gw2 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/docker-compose.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : Stderr: Network roottestovercommittracker-gw2_default Creating (cluster.py:147, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : Stderr: Network roottestovercommittracker-gw2_default Created (cluster.py:147, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Creating (cluster.py:147, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Created (cluster.py:147, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Starting (cluster.py:147, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Started (cluster.py:147, run_and_check) 2026-04-10 18:55:34.056000 [ 638 ] DEBUG : ClickHouse instance created (cluster.py:3157, start) 2026-04-10 18:55:34.057000 [ 638 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2015, get_instance_ip) 2026-04-10 18:55:34.059000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestovercommittracker-gw2-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.059000 [ 638 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 18:55:34.062000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestovercommittracker-gw2-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.063000 [ 638 ] DEBUG : Waiting for ClickHouse start in node, ip: 172.16.1.2... (cluster.py:3165, start) 2026-04-10 18:55:34.065000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestovercommittracker-gw2-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.067000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.171000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.274000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.378000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.481000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.585000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.689000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.792000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.896000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:34.999000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.103000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.206000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.309000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.413000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.517000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.620000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.724000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.828000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:35.931000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.034000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.138000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.241000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.345000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.449000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.552000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.656000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.759000 [ 638 ] DEBUG : http://localhost:None "GET /v1.46/containers/0c62df8c4a550bf4802cf5958eeaf3fb6ea88b20e454078311dd879110c280a1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 18:55:36.761000 [ 638 ] DEBUG : ClickHouse node started (cluster.py:3169, start) ------------------------------ Captured log call ------------------------------- 2026-04-10 18:55:36.764000 [ 638 ] DEBUG : Executing query CREATE USER IF NOT EXISTS A on node (cluster.py:3658, query) 2026-04-10 18:55:37.130000 [ 638 ] DEBUG : Executing query GRANT ALL ON *.* TO A on node (cluster.py:3658, query) 2026-04-10 18:55:37.346000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.347000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.348000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.349000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.350000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.351000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.352000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.353000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.354000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.356000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.357000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.358000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.360000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.362000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.363000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.364000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.375000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.387000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.394000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.407000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.416000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.425000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.437000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.482000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.500000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.524000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.541000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.550000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.572000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.582000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.594000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.618000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.629000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.634000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.648000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.656000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.660000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.673000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.690000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.713000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.730000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.741000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.744000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.769000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.783000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.784000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.799000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.802000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.810000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.822000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.834000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.836000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.859000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.873000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.890000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.952000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.964000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.970000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:37.989000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.003000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.016000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.026000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.028000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.043000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.045000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.046000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.058000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.059000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.081000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.109000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.120000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.135000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.145000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.168000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.180000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.187000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.195000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.202000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.215000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.217000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.230000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.239000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.250000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.255000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.265000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.278000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.280000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.288000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.310000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.314000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.315000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.326000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.337000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.342000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.344000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.357000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.371000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.395000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.410000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=1 on node (cluster.py:3726, get_query_request) 2026-04-10 18:55:38.418000 [ 638 ] DEBUG : Executing query SELECT groupArray(number) FROM numbers(2500000) SETTINGS max_memory_usage_for_user=2000000000,memory_overcommit_ratio_denominator=80000000 on node (cluster.py:3726, get_query_request) ---------------------------- Captured log teardown ----------------------------- 2026-04-10 18:55:58.327000 [ 638 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/.env --project-name roottestovercommittracker-gw2 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/docker-compose.yml stop --timeout 20] (cluster.py:121, run_and_check) 2026-04-10 18:56:03.266000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:56:03.266000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:56:03.266000 [ 638 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 18:56:03.279000 [ 638 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/.env --project-name roottestovercommittracker-gw2 --file /ClickHouse/tests/integration/test_overcommit_tracker/_instances-1-gw2/node/docker-compose.yml down --volumes] (cluster.py:121, run_and_check) 2026-04-10 18:56:03.706000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 18:56:03.706000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 18:56:03.706000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Removing (cluster.py:147, run_and_check) 2026-04-10 18:56:03.707000 [ 638 ] DEBUG : Stderr: Container roottestovercommittracker-gw2-node-1 Removed (cluster.py:147, run_and_check) 2026-04-10 18:56:03.707000 [ 638 ] DEBUG : Stderr: Network roottestovercommittracker-gw2_default Removing (cluster.py:147, run_and_check) 2026-04-10 18:56:03.707000 [ 638 ] DEBUG : Stderr: Network roottestovercommittracker-gw2_default Removed (cluster.py:147, run_and_check) 2026-04-10 18:56:03.707000 [ 638 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 18:56:03.726000 [ 638 ] DEBUG : Docker networks for project roottestovercommittracker-gw2 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 18:56:03.747000 [ 638 ] DEBUG : Docker containers for project roottestovercommittracker-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 18:56:03.768000 [ 638 ] DEBUG : Docker volumes for project roottestovercommittracker-gw2 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 18:56:03.768000 [ 638 ] DEBUG : Command:[docker container list --all --filter name='^/roottestovercommittracker-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 18:56:03.788000 [ 638 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 18:56:03.789000 [ 638 ] DEBUG : No running containers for project: roottestovercommittracker-gw2 (cluster.py:889, cleanup) 2026-04-10 18:56:03.789000 [ 638 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 18:56:03.810000 [ 638 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 18:56:03.810000 [ 638 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 18:56:03.843000 [ 638 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 18:56:03.843000 [ 638 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 18:56:03.843000 [ 638 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 18:56:03.843000 [ 638 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 18:56:03.865000 [ 638 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 18:56:03.865000 [ 638 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) ----------------- generated report log file: parallel1_1.jsonl ----------------- ============================== slowest durations =============================== 21.46s call test_overcommit_tracker/test.py::test_user_overcommit 13.99s setup test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server] 13.96s setup test_overcommit_tracker/test.py::test_user_overcommit 5.54s teardown test_overcommit_tracker/test.py::test_user_overcommit 3.20s teardown test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user] 0.53s call test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user] 0.48s call test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server] 0.47s call test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server] 0.47s call test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user] 0.00s teardown test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server] 0.00s setup test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user] 0.00s setup test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server] 0.00s setup test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user] 0.00s teardown test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server] 0.00s teardown test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user] =========================== short test summary info ============================ FAILED test_overcommit_tracker/test.py::test_user_overcommit - AssertionError... SKIPPED [2] test_max_bytes_ratio_before_external_order_group_by_for_server/test.py:34: TSan build is skipped due to memory overhead SKIPPED [2] test_max_bytes_ratio_before_external_order_group_by_for_server/test.py:65: TSan build is skipped due to memory overhead ======================== 1 failed, 4 skipped in 43.33s ========================= Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 492, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_403iw5 --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[server]' 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_group_by[user]' 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[server]' 'test_max_bytes_ratio_before_external_order_group_by_for_server/test.py::test_max_bytes_ratio_before_external_sort[user]' test_overcommit_tracker/test.py::test_user_overcommit -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 ' returned non-zero exit status 1.